CAR-Net: Clairvoyant Attentive Recurrent Network
نویسندگان
چکیده
We present an interpretable framework for path prediction that learns scene-specific causations behind agents’ behaviors. We exploit two sources of information: the past motion trajectory of the agent of interest and a wide topdown view of the scene. We propose a Clairvoyant Attentive Recurrent Network (CAR-Net) that learns “where to look” in the large image when solving the path prediction task. While previous works on trajectory prediction are constrained to either use semantic information or hand-crafted regions centered around the agent, our method has the capacity to select any region within the image, e.g., a far-away curve when predicting the change of speed of vehicles. To study our goal towards learning observable causality behind agents’ behaviors, we have built a new dataset made of top view images of hundreds of scenes (e.g., F1 racing circuits) where the vehicles are governed by known specific regions within the images (e.g., upcoming curves). Our algorithm successfully selects these regions, learns navigation patterns that generalize to unseen maps, outperforms previous works in terms of prediction accuracy on publicly available datasets, and provides human-interpretable static scene-specific dependencies.
منابع مشابه
Computational design and nonlinear dynamics of recurrent network models of the primary visual cortex
The recurrent neural interaction in the primary visual cortex makes its outputs complex nonlinear functions of its inputs. This nonlinear transform serves the role of pre-attentive visual segmentation, i.e., the autonomous transformation from visual inputs to processed outputs that selectively emphasize certain features (e.g., pop-out features) for segmentation. Understanding the nonlinear dyna...
متن کاملModelling Sentence Pairs with Tree-structured Attentive Encoder
We describe an attentive encoder that combines tree-structured recursive neural networks and sequential recurrent neural networks for modelling sentence pairs. Since existing attentive models exert attention on the sequential structure, we propose a way to incorporate attention into the tree topology. Specially, given a pair of sentences, our attentive encoder uses the representation of one sen...
متن کاملAttentive Language Models
In this paper, we extend Recurrent Neural Network Language Models (RNN-LMs) with an attention mechanism. We show that an Attentive RNN-LM (with 14.5M parameters) achieves a better perplexity than larger RNN-LMs (with 66M parameters) and achieves performance comparable to an ensemble of 10 similar sized RNN-LMs. We also show that an Attentive RNN-LM needs less contextual information to achieve s...
متن کاملIrony Detection with Attentive Recurrent Neural Networks
Automatic Irony Detection refers to making computer understand the real intentions of human behind the ironic language. Much work has been done using classic machine learning techniques applied on various features. In contrast to sophisticated feature engineering, this paper investigates how the deep learning can be applied to the intended task with the help of word embedding. Three different d...
متن کاملAttentive Convolution
In NLP, convolution neural networks (CNNs) have benefited less than recurrent neural networks (RNNs) from attention mechanisms. We hypothesize that this is because attention in CNNs has been mainly implemented as attentive pooling (i.e., it is applied to pooling) rather than as attentive convolution (i.e., it is integrated into convolution). Convolution is the differentiator of CNNs in that it ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1711.10061 شماره
صفحات -
تاریخ انتشار 2017